# Multi-task optimization

Instella 3B Long Instruct
Other
Instella-Long is an open-source language model with 3B parameters developed by AMD, supporting a context length of 128K and performing excellently in long-context benchmark tests.
Large Language Model Transformers
I
amd
240
1
Xgen Small 9B Instruct R
xGen-small is an enterprise-grade compact language model that achieves long-context performance with predictable low costs through domain-focused data curation, scalable pre-training, length extension, and reinforcement learning fine-tuning.
Large Language Model Transformers English
X
Salesforce
97
4
Mmrexcev GRPO V0.420
This is a pre-trained language model merged using the SLERP method, combining the characteristics of both Captain-Eris_Violet-GRPO-v0.420 and MMR-E1 models.
Large Language Model Transformers
M
Nitral-Archive
35
2
Reasoning SCE Coder V1.0
A 32B-parameter large language model constructed based on the SCE fusion method, integrating multiple high-performance pre-trained models
Large Language Model Transformers
R
BenevolenceMessiah
235
3
Labse Ru Sts
MIT
High-quality Russian sentence embedding BERT model, optimized based on cointegrated/LaBSE-en-ru, suitable for semantic text similarity tasks
Text Embedding Transformers Other
L
sergeyzh
4,650
6
Chinese Lert Large
Apache-2.0
LERT is a linguistics theory-driven pre-trained language model designed to enhance model performance through linguistic knowledge.
Large Language Model Transformers Chinese
C
hfl
197
15
Gottbert Base Last
MIT
GottBERT is the first RoBERTa model specifically designed for German, pre-trained on the German portion of the OSCAR dataset, available in both base and large versions.
Large Language Model German
G
TUM
6,842
17
Mengzi Bert Base
Apache-2.0
BERT model pre-trained on 300G Chinese corpus using MLM, POS, and SOP tasks
Large Language Model Transformers Chinese
M
Langboat
438
37
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase